sinc kernel
- North America > United States (0.14)
- South America > Chile (0.04)
- North America > Canada > British Columbia > Metro Vancouver Regional District > Vancouver (0.04)
- Asia > India (0.04)
Band-Limited Gaussian Processes: The Sinc Kernel
We propose a novel class of Gaussian processes (GPs) whose spectra have compact support, meaning that their sample trajectories are almost-surely band limited. As a complement to the growing literature on spectral design of covariance kernels, the core of our proposal is to model power spectral densities through a rectangular function, which results in a kernel based on the sinc function with straightforward extensions to non-centred (around zero frequency) and frequency-varying cases. In addition to its use in regression, the relationship between the sinc kernel and the classic theory is illuminated, in particular, the Shannon-Nyquist theorem is interpreted as posterior reconstruction under the proposed kernel. Additionally, we show that the sinc kernel is instrumental in two fundamental signal processing applications: first, in stereo amplitude modulation, where the non-centred sinc kernel arises naturally. Second, for band-pass filtering, where the proposed kernel allows for a Bayesian treatment that is robust to observation noise and missing data. The developed theory is complemented with illustrative graphic examples and validated experimentally using real-world data.
- South America > Chile (0.04)
- North America > Canada (0.04)
- Asia > India (0.04)
Trainable Time Warping: Aligning Time-Series in the Continuous-Time Domain
Khorram, Soheil, McInnis, Melvin G, Provost, Emily Mower
DTW calculates the similarity or alignment between two signals, subject to temporal warping. However, its computational complexity grows exponentially with the number of time-series. Although there have been algorithms developed that are linear in the number of time-series, they are generally quadratic in time-series length. The exception is generalized time warping (GTW), which has linear computational cost. Yet, it can only identify simple time warping functions. There is a need for a new fast, high-quality multisequence alignment algorithm. We introduce trainable time warping (TTW), whose complexity is linear in both the number and the length of time-series. TTW performs alignment in the continuous-time domain using a sinc convolutional kernel and a gradient-based optimization technique. We compare TTW and GTW on 85 UCR datasets in time-series averaging and classification. TTW outperforms GTW on 67.1% of the datasets for the averaging tasks, and 61.2% of the datasets for the classification tasks.
- North America > United States > Michigan > Washtenaw County > Ann Arbor (0.04)
- North America > United States > California > Riverside County > Riverside (0.04)
Near-Optimal Coresets of Kernel Density Estimates
Phillips, Jeff M., Tai, Wai Ming
We construct near-optimal coresets for kernel density estimate for points in $\mathbb{R^d}$ when the kernel is positive definite. Specifically we show a polynomial time construction for a coreset of size $O(\sqrt{d\log (1/\epsilon)}/\epsilon)$, and we show a near-matching lower bound of size $\Omega(\sqrt{d}/\epsilon)$. The upper bound is a polynomial in $1/\epsilon$ improvement when $d \in [3,1/\epsilon^2)$ (for all kernels except the Gaussian kernel which had a previous upper bound of $O((1/\epsilon) \log^d (1/\epsilon))$) and the lower bound is the first known lower bound to depend on $d$ for this problem. Moreover, the upper bound restriction that the kernel is positive definite is significant in that it applies to a wide-variety of kernels, specifically those most important for machine learning. This includes kernels for information distances and the sinc kernel which can be negative.